AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
128k long context processing

# 128k long context processing

Granite 4.0 Tiny Base Preview
Apache-2.0
Granite-4.0-Tiny-Base-Preview is a 7-billion parameter Mixture of Experts (MoE) language model developed by IBM, featuring a 128k token context window and enhanced expressive capabilities through Mamba-2 technology.
Large Language Model Transformers
G
ibm-granite
156
12
Gemma 3 27B It Qat GGUF
The Gemma 3 27B IT model, introduced by Google, is suitable for various text generation and image understanding tasks, supporting a context length of 128k tokens and multimodal image processing.
Image-to-Text
G
lmstudio-community
41.35k
8
Qwen2.5 QwQ 37B Eureka Triple Cubed
Apache-2.0
An enhanced version of QwQ-32B, improving reasoning and output capabilities through 'cubed' and 'triple-cubed' methods, supporting 128k context.
Large Language Model Transformers Other
Q
DavidAU
210
5
Mistral Nemo Instruct 2407
Apache-2.0
Mistral-Nemo-Instruct-2407 is a large language model fine-tuned on instructions based on Mistral-Nemo-Base-2407, jointly trained by Mistral AI and NVIDIA, outperforming existing models of similar or smaller size.
Large Language Model Transformers Supports Multiple Languages
M
mistralai
149.79k
1,519
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase